Paying Attention to SQuAD: Exploring Bidirectional Attention Flow
نویسندگان
چکیده
With the goal of automated reading comprehension, we apply a neural network with Bidirectional Attention Flow (BiDAF) to the Stanford Question Answering Dataset (SQuAD) and achieve F1 and Exact Match (EM) scores close to the original paper with a single model. We obtain a test F1 score of 76.037 and test EM score of 66.663. Our model includes Character-level CNN embeddings, a Highway Network layer, a Phrase Embedding layer, a Modeling layer, and smart span selection. We also explored expanding the model with feature engineering and an Answer Pointer output layer, which did not further improve our best model. We analyze our model’s performance across categories of contexts, questions, and answers, and compare baseline attention with BiDAF.
منابع مشابه
Bidirectional Attention Flow for Machine Comprehension
Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a ...
متن کاملAttention-based Recurrent Neural Networks for Question Answering
Machine Comprehension (MC) of text is an important problem in Natural Language Processing (NLP) research, and the task of Question Answering (QA) is a major way of assessing MC outcomes. One QA dataset that has gained immense popularity recently is the Stanford Question Answering Dataset (SQuAD). Successful models for SQuAD have all involved the use of Recurrent Neural Network (RNN), and most o...
متن کاملCS224n Assignment 4: Machine Comprehension with Exploration on Attention Mechanism
This goal of this paper is to perform the prediction task on SQuAD dataset about reading comprehension. Given a pair of context paragraph and a question, we’ll output an answer. To do this, a model is built combining the idea of Bidirectional LSTM and attention flow mechanism. The basic architecture and setup details of the model are introduced, so do the summary of performance and error analys...
متن کاملSeq2seq-Attention Question Answering Model
A sequence-to-sequence attention reading comprehension model was implemented to fulfill Question Answering task defined in Stanford Question Answering Dataset (SQuAD). The basic structure was bidirectional LSTM (BiLSTM) encodings with attention mechanism as well as BiLSTM decoding. Several adjustments such as dropout, learning rate decay, and gradients clipping were used. Finally, the model ach...
متن کاملLooking for Leptonic Cp Violation with Neutrinos
I discuss some theoretical aspects of how to observe leptonic CP violation. It is divided into two parts, one for CP violation due to Majorana, and the other more conventional leptonic Kobayashi-Maskawa (KM) phases. In the first part, I estimate the effect of Majorana phase to observable of neutrinoless double beta decay experiments by paying a careful attention to the definition of the atmosph...
متن کامل